Goto

Collaborating Authors

 human perception


Sycophancy Claims about Language Models: The Missing Human-in-the-Loop

Batzner, Jan, Stocker, Volker, Schmid, Stefan, Kasneci, Gjergji

arXiv.org Artificial Intelligence

Sycophantic response patterns in Large Language Models (LLMs) have been increasingly claimed in the literature. We review methodological challenges in measuring LLM sycophancy and identify five core operationalizations. Despite sycophancy being inherently human-centric, current research does not evaluate human perception. Our analysis highlights the difficulties in distinguishing sycophantic responses from related concepts in AI alignment and offers actionable recommendations for future research. Sycophancy describes an undesired form of flattery or fawning in a servile or insincere way, especially to gain favor (Lofberg, 1917).


The Role of Consequential and Functional Sound in Human-Robot Interaction: Toward Audio Augmented Reality Interfaces

Smith, Aliyah, Kennedy, Monroe III

arXiv.org Artificial Intelligence

Abstract--As robots become increasingly integrated into everyday environments, understanding how they communicate with humans is critical. Sound offers a powerful channel for interaction, encompassing both operational noises and intentionally designed auditory cues. In this study, we examined the effects of consequential and functional sounds on human perception and behavior, including a novel exploration of spatial sound through localization and handover tasks. Results show that consequential sounds of the Kinova Gen3 manipulator did not negatively affect perceptions, spatial localization is highly accurate for lateral cues but declines for frontal cues, and spatial sounds can simultaneously convey task-relevant information while promoting warmth and reducing discomfort. These findings highlight the potential of functional and transformative auditory design to enhance human-robot collaboration and inform future sound-based interaction strategies. UDIO Augmented Reality remains a comparatively un-derexplored domain within the broader field of Augmented Reality (AR) research [1]. While recent advancements in AR technologies have spurred extensive investigation into visual augmentation--where virtual objects are seamlessly integrated into the physical environment--research on auditory augmentation has lagged behind.




172ef5a94b4dd0aa120c6878fc29f70c-AuthorFeedback.pdf

Neural Information Processing Systems

We thank the reviewers for encouraging and insightful comments. Below we respond to reviewers' major comments. How does it correlate with human perception? We appreciate this question about the essential motivation of this work. Model attributions are facts about a model's However in the opposite direction, one can never trust a model with brittle attributions.





CLIP is All You Need for Human-like Semantic Representations in Stable Diffusion

Braunstein, Cameron, Toneva, Mariya, Ilg, Eddy

arXiv.org Artificial Intelligence

Latent diffusion models such as Stable Diffusion achieve state-of-the-art results on text-to-image generation tasks. However, the extent to which these models have a semantic understanding of the images they generate is not well understood. In this work, we investigate whether the internal representations used by these models during text-to-image generation contain semantic information that is meaningful to humans. To do so, we perform probing on Stable Diffusion with simple regression layers that predict semantic attributes for objects and evaluate these predictions against human annotations. Surprisingly, we find that this success can actually be attributed to the text encoding occurring in CLIP rather than the reverse diffusion process. We demonstrate that groups of specific semantic attributes have markedly different decoding accuracy than the average, and are thus represented to different degrees. Finally, we show that attributes become more difficult to disambiguate from one another during the inverse diffusion process, further demonstrating the strongest semantic representation of object attributes in CLIP. We conclude that the separately trained CLIP vision-language model is what determines the human-like semantic representation, and that the diffusion process instead takes the role of a visual decoder.


Decoding the Ear: A Framework for Objectifying Expressiveness from Human Preference Through Efficient Alignment

Lin, Zhiyu, Yang, Jingwen, Zhao, Jiale, Liu, Meng, Li, Sunzhu, Wang, Benyou

arXiv.org Artificial Intelligence

Recent speech-to-speech (S2S) models generate intelligible speech but still lack natural expressiveness, largely due to the absence of a reliable evaluation metric. Existing approaches, such as subjective MOS ratings, low-level acoustic features, and emotion recognition are costly, limited, or incomplete. To address this, we present DeEAR (Decoding the Expressive Preference of eAR), a framework that converts human preference for speech expressiveness into an objective score. Grounded in phonetics and psychology, DeEAR evaluates speech across three dimensions: Emotion, Prosody, and Spontaneity, achieving strong alignment with human perception (Spearman's Rank Correlation Coefficient, SRCC = 0.86) using fewer than 500 annotated samples. Beyond reliable scoring, DeEAR enables fair benchmarking and targeted data curation. It not only distinguishes expressiveness gaps across S2S models but also selects 14K expressive utterances to form ExpressiveSpeech, which improves the expressive score (from 2.0 to 23.4 on a 100-point scale) of S2S models. Demos and codes are available at https://github.com/FreedomIntelligence/ExpressiveSpeech